skip to main content
US FlagAn official website of the United States government
dot gov icon
Official websites use .gov
A .gov website belongs to an official government organization in the United States.
https lock icon
Secure .gov websites use HTTPS
A lock ( lock ) or https:// means you've safely connected to the .gov website. Share sensitive information only on official, secure websites.


Search for: All records

Creators/Authors contains: "Su, Yu"

Note: When clicking on a Digital Object Identifier (DOI) number, you will be taken to an external site maintained by the publisher. Some full text articles may not yet be available without a charge during the embargo (administrative interval).
What is a DOI Number?

Some links on this page may take you to non-federal websites. Their policies may differ from this site.

  1. Free, publicly-accessible full text available April 25, 2026
  2. We present a simple approach to make pre-trained Vision Transformers (ViTs) interpretable for fine-grained analysis, aiming to identify and localize the traits that distinguish visually similar categories, such as bird species. Pre-trained ViTs, such as DINO, have demonstrated remarkable capabilities in extracting localized, discriminative features. However, saliency maps like Grad-CAM often fail to identify these traits, producing blurred, coarse heatmaps that highlight entire objects instead. We propose a novel approach, Prompt Class Attention Map (Prompt-CAM), to address this limitation. Prompt-CAM learns class-specific prompts for a pre-trained ViT and uses the corresponding outputs for classification. To correctly classify an image, the true-class prompt must attend to unique image patches not present in other classes' images (i.e., traits). As a result, the true class's multi-head attention maps reveal traits and their locations. Implementation-wise, Prompt-CAM is almost a "free lunch," requiring only a modification to the prediction head of Visual Prompt Tuning (VPT). This makes Prompt-CAM easy to train and apply, in stark contrast to other interpretable methods that require designing specific models and training processes. Extensive empirical studies on a dozen datasets from various domains (e.g., birds, fishes, insects, fungi, flowers, food, and cars) validate the superior interpretation capability of Prompt-CAM. The source code and demo are available at https://github.com/Imageomics/Prompt_CAM. 
    more » « less
    Free, publicly-accessible full text available June 1, 2026
  3. Free, publicly-accessible full text available December 12, 2025
  4. Free, publicly-accessible full text available December 10, 2025
  5. Free, publicly-accessible full text available December 12, 2025
  6. We investigate the quantum dynamics of a spin coupling to a bath of independent spins via the dissipaton equation of motion (DEOM) approach. The bath, characterized by a continuous spectral density function, is composed of spins that are independent level systems described by the su(2) Lie algebra, representing an environment with a large magnitude of anharmonicity. Based on the previous work by Suarez and Silbey [J. Chem. Phys. 95, 9115 (1991)] and by Makri [J. Chem. Phys. 111, 6164 (1999)] that the spin bath can be mapped to a Gaussian environment under its linear response limit, we use the time-domain Prony fitting decomposition scheme to the bare–bath time correlation function (TCF) given by the bosonic fluctuation–dissipation theorem to generate the exponential decay basis (or pseudo modes) for DEOM construction. The accuracy and efficiency of this strategy have been explored by a variety of numerical results. We envision that this work provides new insights into extending the hierarchical equations of motion and DEOM approach to certain types of anharmonic environments with arbitrary TCF or spectral density. 
    more » « less
  7. Free, publicly-accessible full text available April 25, 2026
  8. Free, publicly-accessible full text available December 12, 2025